Goto

Collaborating Authors

 Web


Data-Dependent Bounds for Online Portfolio Selection Without Lipschitzness and Smoothness Chung-En Tsai Department of Computer Science and Information Engineering National Taiwan University

Neural Information Processing Systems

This work introduces the first small-loss and gradual-variation regret bounds for online portfolio selection, marking the first instances of data-dependent bounds for online convex optimization with non-Lipschitz, non-smooth losses. The algorithms we propose exhibit sublinear regret rates in the worst cases and achieve logarithmic regrets when the data is "easy," with per-round time almost linear in the number of investment alternatives. The regret bounds are derived using novel smoothness characterizations of the logarithmic loss, a local norm-based analysis of following the regularized leader (FTRL) with self-concordant regularizers, which are not necessarily barriers, and an implicit variant of optimistic FTRL with the log-barrier.




This Google Chrome update could change the fundamentals of browsing - here's who gets to try it first

ZDNet

Google's Chrome browser for MacOS and Windows is receiving an infusion of new Gemini-powered capabilities, including an AI browsing assistant contextually sensitized to a user's browsing activities. Google made the announcement this week at Google I/O 2025. Dubbed Gemini-in-Chrome, the feature will be available May 21 to Google AI Pro and Google AI Ultra subscribers in the US as well as Chrome Beta, Dev, and Canary users. The general idea behind Gemini-in-Chrome is to reorganize, aggregate, and then more sensibly redisplay the data found on one or more browser tabs while also embellishing the final output with additional but relevant Gemini-generated information. For example, during a pre-event press briefing attended by ZDNET, Google director of Chrome product management Charmaine D'Silva demonstrated how Gemini-in-Chrome could not only organize a head-to-head feature comparison chart of individual sleeping bags -- to which multiple Chrome tabs (one tab per sleeping bag) were pointing -- but could respond to text prompts about each bag's suitability to the expected temperatures for an upcoming camping trip in Maine.


Google's 'AI Mode' search is ready to replace a list of links

PCWorld

Google said Thursday that it has begun migrating its "AI Mode" out of its experimental Labs effort and into the real world. Google said that a "small percentage of people" in the "coming weeks" will see what Google calls AI Mode, or entirely AI-generated responses to queries that users ask. It's Google's response to services like Anthropic, which "answer" queries using AI, which slurps up and regurgitates answers that others have already provided. Google first began revamping its search algorithm in 2023, when it started aggregating AI-powered summaries of say, the best laptops. AI has been used elsewhere by Google services like Chrome to sum up web pages, as well.


Pile of Law: Learning Responsible Data Filtering from the Law and a 256GB Open-Source Legal Dataset

Neural Information Processing Systems

Emerging ethical approaches have attempted to filter pretraining material, but such approaches have been ad hoc and failed to take context into account. We offer an approach to filtering grounded in law, which has directly addressed the tradeoffs in filtering material. First, we gather and make available the Pile of Law, a 256GB (and growing) dataset of open-source English-language legal and administrative data, covering court opinions, contracts, administrative rules, and legislative records. Pretraining on the Pile of Law may help with legal tasks that have the promise to improve access to justice. Second, we distill the legal norms that governments have developed to constrain the inclusion of toxic or private content into actionable lessons for researchers and discuss how our dataset reflects these norms. Third, we show how the Pile of Law offers researchers the opportunity to learn such filtering rules directly from the data, providing an exciting new research direction in model-based processing. Warning: this paper contains quotations that may be offensive or upsetting.


Consent in Crisis: The Rapid Decline of the AI Data Commons, Ariel Lee

Neural Information Processing Systems

General-purpose artificial intelligence (AI) systems are built on massive swathes of public web data, assembled into corpora such as C4, RefinedWeb, and Dolma. To our knowledge, we conduct the first, large-scale, longitudinal audit of the consent protocols for the web domains underlying AI training corpora. Our audit of 14, 000 web domains provides an expansive view of crawlable web data and how codified data use preferences are changing over time. We observe a proliferation of AIspecific clauses to limit use, acute differences in restrictions on AI developers, as well as general inconsistencies between websites' expressed intentions in their Terms of Service and their robots.txt. We diagnose these as symptoms of ineffective web protocols, not designed to cope with the widespread re-purposing of the internet for AI.



On the Effects of Data Scale on UI Control Agents

Neural Information Processing Systems

Autonomous agents that control user interfaces to accomplish human tasks are emerging. Leveraging LLMs to power such agents has been of special interest, but unless fine-tuned on human-collected task demonstrations, performance is still relatively low. In this work we study whether fine-tuning alone is a viable approach for building real-world UI control agents.


Synatra: Turning Indirect Knowledge into Direct Demonstrations for Digital Agents at Scale

Neural Information Processing Systems

LLMs can now act as autonomous agents that interact with digital environments and complete specific objectives (e.g., arranging an online meeting). However, accuracy is still far from satisfactory, partly due to a lack of large-scale, direct demonstrations for digital tasks. Obtaining supervised data from humans is costly, and automatic data collection through exploration or reinforcement learning relies on complex environmental and content setup, resulting in datasets that lack comprehensive coverage of various scenarios. On the other hand, there is abundant knowledge that may indirectly assist task completion, such as online tutorials that were created for human consumption. In this work, we present Synatra, an approach that effectively transforms this indirect knowledge into direct supervision at scale. We define different types of indirect knowledge, and carefully study the available sources to obtain it, methods to encode the structure of direct demonstrations, and finally methods to transform indirect knowledge into direct demonstrations. We use 100k such synthetically-created demonstrations to finetune a 7B CodeLlama, and demonstrate that the resulting agent surpasses all comparably sized models on three web-based task benchmarks Mind2Web, MiniWoB++ and WebArena, as well as surpassing GPT-3.5 on WebArena and Mind2Web. In addition, while synthetic demonstrations prove to be only 3% the cost of human demonstrations (at $0.031 each), we show that the synthetic demonstrations can be more effective than an identical number of human demonstrations collected from limited domains.